Energy-Efficient Neural Networks using Approximate Computation Reuse
نویسندگان
چکیده
As a problem-solving method, neural networks have shown broad success for medical applications, speech recognition, and natural language processing. Current hardware implementations of neural networks exhibit high energy consumption due to the intensive computing workloads. This paper proposes a methodology to design an energy-efficient neural network that effectively exploits computation reuse opportunities. To do so, we use Bloom filters (BFs) by tightly integrating them with computation units. BFs store and recall frequently occurring input patterns to reuse computations. We expand the opportunities for computation reuse by storing frequent input patterns specific to a given layer and using approximate pattern matching with hashing for limited data precision. This reconfigurable matching is key to achieving a “controllable approximation” for neural networks. To lower the energy consumption of BFs, we also use low-pow memristor arrays to implement BFs. Our experimental results show that for convolutional neural networks, the BFs enable 47.5% energy saving of multiplication operations, while incurring only 1% accuracy drop. While the actual savings will vary depending upon the extent of approximation and reuse, this paper presents a method for reducing computing workloads and improving energy efficiency.
منابع مشابه
Computation of the linear Schrodinger Energy levels by Sinc method
Computation of the Schrodinger equation energy levels is very important in physics. For example we can use these levels to calculate the absorption coefficient and light refraction in a material as well as calculation of the interband and intersubband transition energies. Density of states of the system can also be obtained by using the energy levels. Thereafter we can determine that the system...
متن کاملA Multithreaded CGRA for Convolutional Neural Network Processing
Convolutional neural network (CNN) is an essential model to achieve high accuracy in various machine learning applications, such as image recognition and natural language processing. One of the important issues for CNN acceleration with high energy efficiency and processing performance is efficient data reuse by exploiting the inherent data locality. In this paper, we propose a novel CGRA (Coar...
متن کاملA Comparative Approximate Economic Behavior Analysis Of Support Vector Machines And Neural Networks Models
متن کامل
Escort: Efficient Sparse Convolutional Neural Networks on GPUs
Deep neural networks have achieved remarkable accuracy in many artificial intelligence applications, e.g. computer vision, at the cost of a large number of parameters and high computational complexity. Weight pruning can compress DNN models by removing redundant parameters in the networks, but it brings sparsity in the weight matrix, and therefore makes the computation inefficient on GPUs. Alth...
متن کاملApproximate Computation using Neuralized FPU
Neural networks can be used as function approximators to improve the energy efficiency, performance, and fault-tolerance of traditional computer architectures. To maximize these improvements the granularity of the function must be as large as possible. This work-inprogress abstract explores the lower limits of neural network function approximation by replacing individual floating point multipli...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017